Hardness results for neural network approximation problems
نویسندگان
چکیده
منابع مشابه
Approximation and Hardness Results for Label Cut and Related Problems
We investigate a natural combinatorial optimization problem called the Label Cut problem. Given an input graph G with a source s and a sink t, the edges of G are classified into different categories, represented by a set of labels. The labels may also have weights. We want to pick a subset of labels of minimum cardinality (or minimum total weight), such that the removal of all edges with these ...
متن کاملApproximation Algorithms and Hardness Results for Labeled Connectivity Problems
Let G = (V, E) be a connected multigraph, whose edges are associated with labels specified by an integer-valued function L : E → N. In addition, each label ` ∈ N has a non-negative cost c(`). The minimum label spanning tree problem (MinLST) asks to find a spanning tree in G that minimizes the overall cost of the labels used by its edges. Equivalently, we aim at finding a minimum cost subset of ...
متن کاملHardness of Approximation for Vertex-Connectivity Network-Design Problems
In the survivable network design problem (SNDP), the goal is to find a minimum-cost spanning subgraph satisfying certain connectivity requirements. We study the vertex-connectivity variant of SNDP in which the input specifies, for each pair of vertices, a required number of vertex-disjoint paths connecting them. We give the first strong lower bound on the approximability of SNDP, showing that t...
متن کاملDegree-Constrained Subgraph Problems: Hardness and Approximation Results
A general instance of a Degree-Constrained Subgraph problem consists of an edge-weighted or vertex-weighted graph G and the objective is to find an optimal weighted subgraph, subject to certain degree constraints on the vertices of the subgraph. This paper considers two natural Degree-Constrained Subgraph problems and studies their behavior in terms of approximation algorithms. These problems t...
متن کاملHardness results for neural networkapproximation
We consider the problem of eeciently learning in two-layer neural networks. We investigate the computational complexity of agnostically learning with simple families of neural networks as the hypothesis classes. We show that it is NP-hard to nd a linear threshold network of a xed size that approximatelyminimizes the proportion of misclassiied examples in a training set, even if there is a netwo...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Theoretical Computer Science
سال: 2002
ISSN: 0304-3975
DOI: 10.1016/s0304-3975(01)00057-3